Goto

Collaborating Authors

 collect data


Generative AI is learning to spy for the US military

MIT Technology Review

"We still need to validate the sources," says Lowdon. But the unit's commanders encouraged the use of large language models, he says, "because they provide a lot more efficiency during a dynamic situation." The generative AI tools they used were built by the defense-tech company Vannevar Labs, which in November was granted a production contract worth up to 99 million by the Pentagon's startup-oriented Defense Innovation Unit with the goal of bringing its intelligence tech to more military units. The company, founded in 2019 by veterans of the CIA and US intelligence community, joins the likes of Palantir, Anduril, and Scale AI as a major beneficiary of the US military's embrace of artificial intelligence--not only for physical technologies like drones and autonomous vehicles but also for software that is revolutionizing how the Pentagon collects, manages, and interprets data for warfare and surveillance. Though the US military has been developing computer vision models and similar AI tools, like those used in Project Maven, since 2017, the use of generative AI--tools that can engage in human-like conversation like those built by Vannevar Labs--represent a newer frontier.


Online Location Planning for AI-Defined Vehicles: Optimizing Joint Tasks of Order Serving and Spatio-Temporal Heterogeneous Model Fine-Tuning

Zheng, Bokeng, Rao, Bo, Zhu, Tianxiang, Tan, Chee Wei, Duan, Jingpu, Zhou, Zhi, Chen, Xu, Zhang, Xiaoxi

arXiv.org Artificial Intelligence

Abstract--Advances in artificial intelligence (AI) including foundation models (FMs), are increasingly transforming human society, with smart city driving the evolution of urban living. Meanwhile, vehicle crowdsensing (VCS) has emerged as a key enabler, leveraging vehicles' mobility and sensor-equipped capabilities. In particular, ride-hailing vehicles can effectively facilitate flexible data collection and contribute towards urban intelligence, despite resource limitations. Therefore, this work explores a promising scenario, where edge-assisted vehicles perform joint tasks of order serving and the emerging foundation model finetuning using various urban data. However, integrating the VCS AI task with the conventional order serving task is challenging, due to their inconsistent spatio-temporal characteristics: (i) The distributions of ride orders and data point-of-interests (PoIs) may not coincide in geography, both following a priori unknown patterns; (ii) they have distinct forms of temporal effects, i.e., prolonged waiting makes orders become instantly invalid while data with increased staleness gradually reduces its utility for model fine-tuning. To overcome these obstacles, we propose an online framework based on multi-agent reinforcement learning (MARL) with careful augmentation. A new quality-of-service (QoS) metric is designed to characterize and balance the utility of the two joint tasks, under the effects of varying data volumes and staleness. Each RSU, equipped with a server, stores a complete base model, enabling vehicles to perform real-time fine-tuning as they collect data and transfer the I. X. Zhang are with the School of Computer Science and A previous version appears at IWQoS 2024 as a short paper. Due to the large volume, data stored in the government agencies in better city management. Notably, ridehailing RSU server can be discarded in a certain period of time. In vehicles are particularly advantageous for VCS tasks, practice, these data can be descriptive features and feedbacks due to their centralized ride-hailing platform management, (labels) of recommendation or generative AR applications, which reduces the cost of deploying and executing crowdsensing generated by nearby visitors or residents. They can also be tasks, and utilizes the data and computing resources traffic/environment monitoring data with labels generated by from ride-hailing vehicles to maximize the VCS task utilities. The government or any company that collaborates model (FM)-powered AI applications have revolutionized with the ride-hailing vehicle company has multiple types of numerous aspects of human lives, including healthcare, education, VSC tasks to fulfill, each of which needs certain locations industry, etc. FMs, e.g., BERT, GPT-4, ViT, serve of data for fine-tuning UFMs.


Embodied Active Learning of Generative Sensor-Object Models

Pinosky, Allison, Murphey, Todd D.

arXiv.org Artificial Intelligence

When a robot encounters a novel object, how should it respond$\unicode{x2014}$what data should it collect$\unicode{x2014}$so that it can find the object in the future? In this work, we present a method for learning image features of an unknown number of novel objects. To do this, we use active coverage with respect to latent uncertainties of the novel descriptions. We apply ergodic stability and PAC-Bayes theory to extend statistical guarantees for VAEs to embodied agents. We demonstrate the method in hardware with a robotic arm; the pipeline is also implemented in a simulated environment. Algorithms and simulation are available open source, see http://sites.google.com/u.northwestern.edu/embodied-learning-hardware .


ARCap: Collecting High-quality Human Demonstrations for Robot Learning with Augmented Reality Feedback

Chen, Sirui, Wang, Chen, Nguyen, Kaden, Fei-Fei, Li, Liu, C. Karen

arXiv.org Artificial Intelligence

Recent progress in imitation learning from human demonstrations has shown promising results in teaching robots manipulation skills. To further scale up training datasets, recent works start to use portable data collection devices without the need for physical robot hardware. However, due to the absence of on-robot feedback during data collection, the data quality depends heavily on user expertise, and many devices are limited to specific robot embodiments. We propose ARCap, a portable data collection system that provides visual feedback through augmented reality (AR) and haptic warnings to guide users in collecting high-quality demonstrations. Through extensive user studies, we show that ARCap enables novice users to collect robot-executable data that matches robot kinematics and avoids collisions with the scenes. With data collected from ARCap, robots can perform challenging tasks, such as manipulation in cluttered environments and long-horizon cross-embodiment manipulation. ARCap is fully open-source and easy to calibrate; all components are built from off-the-shelf products. More details and results can be found on our website: https://stanford-tml.github.io/ARCap


Bats' weird wings inspired this drone

Popular Science

Bats are amongst the animal kingdom's most unorthodox fliers. Unlike birds, the furry, flying mammals can dynamically reshape and morph their wings to achieve maximum force and hover in place. The soft membrane of their wings, which more closely resembles a human arm than a bird's wing, is also extremely flexible, which means bats can contour themselves to squeeze into tiny corridors. Now, researchers from Northeastern University are leaning on those unique elements and applying them to a fully autonomous flying drone called "Aerobat." Eventually, they believe this bat-inspired robot could be used to navigate sewer tunnels, caves, and other tight corridors largely off-limits to current flying robots.


A Data Efficient Framework for Learning Local Heuristics

Veerapaneni, Rishi, Park, Jonathan, Saleem, Muhammad Suhail, Likhachev, Maxim

arXiv.org Artificial Intelligence

With the advent of machine learning, there have been several recent attempts to learn effective and generalizable heuristics. Local Heuristic A* (LoHA*) is one recent method that instead of learning the entire heuristic estimate, learns a "local" residual heuristic that estimates the cost to escape a region (Veerapaneni et al 2023). LoHA*, like other supervised learning methods, collects a dataset of target values by querying an oracle on many planning problems (in this case, local planning problems). This data collection process can become slow as the size of the local region increases or if the domain requires expensive collision checks. Our main insight is that when an A* search solves a start-goal planning problem it inherently ends up solving multiple local planning problems. We exploit this observation to propose an efficient data collection framework that does <1/10th the amount of work (measured by expansions) to collect the same amount of data in comparison to baselines. This idea also enables us to run LoHA* in an online manner where we can iteratively collect data and improve our model while solving relevant start-goal tasks. We demonstrate the performance of our data collection and online framework on a 4D $(x, y, \theta, v)$ navigation domain.


Meta Reinforcement Learning for Strategic IoT Deployments Coverage in Disaster-Response UAV Swarms

Dhuheir, Marwan, Erbad, Aiman, Al-Fuqaha, Ala

arXiv.org Artificial Intelligence

In the past decade, Unmanned Aerial Vehicles (UAVs) have grabbed the attention of researchers in academia and industry for their potential use in critical emergency applications, such as providing wireless services to ground users and collecting data from areas affected by disasters, due to their advantages in terms of maneuverability and movement flexibility. The UAVs' limited resources, energy budget, and strict mission completion time have posed challenges in adopting UAVs for these applications. Our system model considers a UAV swarm that navigates an area collecting data from ground IoT devices focusing on providing better service for strategic locations and allowing UAVs to join and leave the swarm (e.g., for recharging) in a dynamic way. In this work, we introduce an optimization model with the aim of minimizing the total energy consumption and provide the optimal path planning of UAVs under the constraints of minimum completion time and transmit power. The formulated optimization is NP-hard making it not applicable for real-time decision making. Therefore, we introduce a light-weight meta-reinforcement learning solution that can also cope with sudden changes in the environment through fast convergence. We conduct extensive simulations and compare our approach to three state-of-the-art learning models. Our simulation results prove that our introduced approach is better than the three state-of-the-art algorithms in providing coverage to strategic locations with fast convergence.


Google now admits it could collect data in Chrome's Incognito mode

Engadget

When users open an Incognito browser on Chrome, they'll see a notification warning them that other people using their device won't be able to see their activity but that their downloads, bookmarks and reading items will still be saved. Now, Google has updated that disclaimer in Chrome's experimental Canary channel, shortly after agreeing to settle a 5 billion lawsuit accusing it of tracking Incognito users. As first noticed by MSPowerUser, the company has tweaked the disclaimer in Canary to add language that says Incognito mode won't change how websites collect people's data. "Others who use this device won't see your activity, so you can browse more privately," the new disclaimer reads. "This won't change how data is collected by websites you visit and the services they use, including Google. Downloads, bookmarks and reading list items will be saved."


Model-aided Federated Reinforcement Learning for Multi-UAV Trajectory Planning in IoT Networks

Chen, Jichao, Esrafilian, Omid, Bayerlein, Harald, Gesbert, David, Caccamo, Marco

arXiv.org Artificial Intelligence

Deploying teams of unmanned aerial vehicles (UAVs) to harvest data from distributed Internet of Things (IoT) devices requires efficient trajectory planning and coordination algorithms. Multi-agent reinforcement learning (MARL) has emerged as a solution, but requires extensive and costly real-world training data. To tackle this challenge, we propose a novel model-aided federated MARL algorithm to coordinate multiple UAVs on a data harvesting mission with only limited knowledge about the environment. The proposed algorithm alternates between building an environment simulation model from real-world measurements, specifically learning the radio channel characteristics and estimating unknown IoT device positions, and federated QMIX training in the simulated environment. Each UAV agent trains a local QMIX model in its simulated environment and continuously consolidates it through federated learning with other agents, accelerating the learning process. A performance comparison with standard MARL algorithms demonstrates that our proposed model-aided FedQMIX algorithm reduces the need for real-world training experiences by around three magnitudes while attaining similar data collection performance.


Robot SHARK is deployed in London's Thames river that can collect 1,100lbs of rubbish a DAY

Daily Mail - Science & tech

A robotic shark hungry for plastic is to snap up waste in the River Thames as part of efforts to tackle water pollution. WasteShark is the first marine robot to take London's river by storm, with the ability to'eat' up to 1,100lbs of waste everyday - equivalent to 22,700 plastic bottles. The electric shark has been released in Canary Wharf where it can travel through 3.1 miles (5km) of water before needing a recharge. It comes at a time when plastic waste has almost doubled globally since 2000, with only nine per cent of this successfully recycled, according to an Organisation for Economic Co-operation and Development report. But Britvic-owned Aqua Libra, which is launching the shark, hope to combat this by recycling the collected rubbish wherever possible.